vslam system
Research on visual simultaneous localization and mapping technology based on near infrared light
Ma, Rui, Liu, Mengfang, Li, Boliang, Li, Xinghui
SLAM originated from the probabilistic SLAM problem at the IEEE Robot and Automation Conference held in San Francisco in 1986[2], and experienced three stages of initial theoretical exploration (1986-2004), algorithmic framework development (2004-2015), and system robustness improvement (2015-now)[3]. According to the sensor classification, the SLAM technology can be divided into laser SLAM, visual SLAM, and multi-sensor fusion SLAM. Laser SLAM is scanned by lidar, who are suitable for indoor environment but inaccurate positioning in a single repeated environment[4-6]. Visual SLAM captures images through the camera, acquires positions and maps through image pixels and features, and is suitable for textured rich scenes. In addition, visual SLAM has the advantages of low cost and small size, which can provide intuitive visual input[7-9].
- North America > United States > California > San Francisco County > San Francisco (0.24)
- Europe > Switzerland > Zürich > Zürich (0.14)
- Asia > China > Guangdong Province > Shenzhen (0.04)
- (7 more...)
Drift-free Visual SLAM using Digital Twins
Merat, Roxane, Cioffi, Giovanni, Bauersfeld, Leonard, Scaramuzza, Davide
Globally-consistent localization in urban environments is crucial for autonomous systems such as self-driving vehicles and drones, as well as assistive technologies for visually impaired people. Traditional Visual-Inertial Odometry (VIO) and Visual Simultaneous Localization and Mapping (VSLAM) methods, though adequate for local pose estimation, suffer from drift in the long term due to reliance on local sensor data. While GPS counteracts this drift, it is unavailable indoors and often unreliable in urban areas. An alternative is to localize the camera to an existing 3D map using visual-feature matching. This can provide centimeter-level accurate localization but is limited by the visual similarities between the current view and the map. This paper introduces a novel approach that achieves accurate and globally-consistent localization by aligning the sparse 3D point cloud generated by the VIO/VSLAM system to a digital twin using point-to-plane matching; no visual data association is needed. The proposed method provides a 6-DoF global measurement tightly integrated into the VIO/VSLAM system. Experiments run on a high-fidelity GPS simulator and real-world data collected from a drone demonstrate that our approach outperforms state-of-the-art VIO-GPS systems and offers superior robustness against viewpoint changes compared to the state-of-the-art Visual SLAM systems.
- Information Technology > Artificial Intelligence > Vision (1.00)
- Information Technology > Artificial Intelligence > Robots > Autonomous Vehicles > Drones (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning (1.00)
QueensCAMP: an RGB-D dataset for robust Visual SLAM
Bruno, Hudson M. S., Colombini, Esther L., Givigi, Sidney N. Jr
Visual Simultaneous Localization and Mapping (VSLAM) is a fundamental technology for robotics applications. While VSLAM research has achieved significant advancements, its robustness under challenging situations, such as poor lighting, dynamic environments, motion blur, and sensor failures, remains a challenging issue. To address these challenges, we introduce a novel RGB-D dataset designed for evaluating the robustness of VSLAM systems. The dataset comprises real-world indoor scenes with dynamic objects, motion blur, and varying illumination, as well as emulated camera failures, including lens dirt, condensation, underexposure, and overexposure. Additionally, we offer open-source scripts for injecting camera failures into any images, enabling further customization by the research community. Our experiments demonstrate that ORB-SLAM2, a traditional VSLAM algorithm, and TartanVO, a Deep Learning-based VO algorithm, can experience performance degradation under these challenging conditions. Therefore, this dataset and the camera failure open-source tools provide a valuable resource for developing more robust VSLAM systems capable of handling real-world challenges.
- South America > Brazil > São Paulo > Campinas (0.04)
- North America > United States > Michigan (0.04)
- North America > Canada > Ontario > Kingston (0.04)
Towards Real-Time Gaussian Splatting: Accelerating 3DGS through Photometric SLAM
Hu, Yan Song, Mao, Dayou, Chen, Yuhao, Zelek, John
Abstract-- Initial applications of 3D Gaussian Splatting (3DGS) in Visual Simultaneous Localization and Mapping (VS-LAM) demonstrate the generation of high-quality volumetric reconstructions from monocular video streams. However, despite these promising advancements, current 3DGS integrations have reduced tracking performance and lower operating speeds compared to traditional VSLAM. To address these issues, we propose integrating 3DGS with Direct Sparse Odometry, a monocular photometric SLAM system. We have done preliminary experiments showing that using Direct Sparse Odometry point cloud outputs, as opposed to standard structure-frommotion methods, significantly shortens the training time needed to achieve high-quality renders. I. INTRODUCTION Visual Simultaneous Localization and Mapping (VSLAM) fixed set of poses and initial points from a structure-frommotion is crucial for developing robust mobile robotics. After converting the VSLAM system would reconstruct environments with photorealistic initial points into 3D Gaussians, their positions, sizes, and accuracy from live video input.
Visual SLAM: What are the Current Trends and What to Expect?
Tourani, Ali, Bavle, Hriday, Sanchez-Lopez, Jose Luis, Voos, Holger
Vision-based sensors have shown significant performance, accuracy, and efficiency gain in Simultaneous Localization and Mapping (SLAM) systems in recent years. In this regard, Visual Simultaneous Localization and Mapping (VSLAM) methods refer to the SLAM approaches that employ cameras for pose estimation and map generation. We can see many research works that demonstrated VSLAMs can outperform traditional methods, which rely only on a particular sensor, such as a Lidar, even with lower costs. VSLAM approaches utilize different camera types (e.g., monocular, stereo, and RGB-D), have been tested on various datasets (e.g., KITTI, TUM RGB-D, and EuRoC) and in dissimilar environments (e.g., indoors and outdoors), and employ multiple algorithms and methodologies to have a better understanding of the environment. The mentioned variations have made this topic popular for researchers and resulted in a wide range of VSLAMs methodologies. In this regard, the primary intent of this survey is to present the recent advances in VSLAM systems, along with discussing the existing challenges and trends. We have given an in-depth literature survey of forty-five impactful papers published in the domain of VSLAMs. We have classified these manuscripts by different characteristics, including the novelty domain, objectives, employed algorithms, and semantic level. We also discuss the current trends and future directions that may help researchers investigate them.
- Europe > Luxembourg (0.14)
- Europe > United Kingdom > England > Greater London > London (0.04)
- Europe > Switzerland > Zürich > Zürich (0.04)
- Europe > Ireland (0.04)
- Research Report (1.00)
- Overview (1.00)